On Proving Linear Convergence of Comparison-based Step-size Adaptive Randomized Search on Scaling-Invariant Functions via Stability of Markov Chains

نویسندگان

  • Anne Auger
  • Nikolaus Hansen
چکیده

In the context of numerical optimization, this paper develops a methodology to analyze the linear convergence of comparison-based step-size adaptive randomized search (CB-SARS), a class of probabilistic derivative-free optimization algorithms where the function is solely used through comparisons of candidate solutions. Various algorithms are included in the class of CB-SARS algorithms. On the one hand, a few methods introduced already in the 60’s: the step-size adaptive random search by Schumer and Steiglitz, the compound random search by Devroye and simplified versions of Matyas’ random optimization algorithm or Kjellstrom and Taxen Gaussian adaptation. On the other hand, it includes simplified versions of several recent algorithms: the covariance-matrix-adaptation evolution strategy algorithm (CMA-ES), the exponential natural evolution strategy (xNES), or the cross entropy method. CB-SARS algorithms typically exhibit several invariances. First of all, invariance to composing the objective function with a strictly monotonic transformation which is a direct consequence of the fact that the algorithms only use comparisons. Second, scale invariance that translates the fact that the algorithm has no intrinsic absolute notion of scale. The algorithms are investigated on scaling-invariant functions defined as functions that preserve the ordering of two points (with respect to their objective function values) when they are both scaled with the same positive parameter (this scaling is done w.r.t. a single constant reference point). This class of functions includes composite of norms by strictly increasing functions as well as some non quasi-convex functions and non-continuous functions. We prove that the algorithm invariance properties entail on scaling-invariant functions the existence of a homogeneous Markov chain whose stability implies linear convergence of the original algorithm. Hence we turn the problem of studying linear convergence of CB-SARS optimization algorithms into a study of the stability of a joint Markov chain. We naturally derive sufficient conditions expressed in terms of different stability conditions for the Markov chain (irreducibility, positivity, Harris recurrence, geometric ergodicity) for the global linear convergence of CB-SARS. The methodology presented here is applied in a companion paper where stability sufficient conditions are proven for the (1+1)-ES with a generalized one-fifth success rule that coincides with Schumer and Steiglitz’ or Devroye’s algorithms. This leads to a first proof of linear convergence for the algorithm considered and first proof of linear convergence for a scale-invariant CB-SARS for such a wide family of functions including in particular non quasi-convex and non continuous functions.

برای دانلود رایگان متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

My title

In this paper, we consider comparison-based adaptive stochastic algorithms for solving numerical optimisation problems. We consider a specific subclass of algorithms that we call comparison-based step-size adaptive randomized search (CB-SARS), where the state variables at a given iteration are a vector of the search space and a positive parameter, the step-size, typically controlling the overal...

متن کامل

Linear Convergence on Positively Homogeneous Functions of a Comparison Based Step-Size Adaptive Randomized Search: the (1+1) ES with Generalized One-fifth Success Rule

In the context of unconstraint numerical optimization, this paper investigates the global linear convergence of a simple probabilistic derivative-free optimization algorithm (DFO). The algorithm samples a candidate solution from a standard multivariate normal distribution scaled by a step-size and centered in the current solution. This solution is accepted if it has a better objective function ...

متن کامل

Markov Chain Analysis of Cumulative Step-Size Adaptation on a Linear Constrained Problem

This paper analyzes a (1, λ)-Evolution Strategy, a randomized comparison-based adaptive search algorithm optimizing a linear function with a linear constraint. The algorithm uses resampling to handle the constraint. Two cases are investigated: first, the case where the step-size is constant, and second, the case where the step-size is adapted using cumulative step-size adaptation. We exhibit fo...

متن کامل

Stability of Adversarial Markov Chains, with an Application to Adaptive MCMC Algorithms

We consider whether ergodic Markov chains with bounded step size remain bounded in probability when their transitions are modified by an adversary on a bounded subset. We provide counterexamples to show that the answer is no in general, and prove theorems to show that the answer is yes under various additional assumptions. We then use our results to prove convergence of various adaptive Markov ...

متن کامل

A Differential Evolution and Spatial Distribution based Local Search for Training Fuzzy Wavelet Neural Network

Abstract   Many parameter-tuning algorithms have been proposed for training Fuzzy Wavelet Neural Networks (FWNNs). Absence of appropriate structure, convergence to local optima and low speed in learning algorithms are deficiencies of FWNNs in previous studies. In this paper, a Memetic Algorithm (MA) is introduced to train FWNN for addressing aforementioned learning lacks. Differential Evolution...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

عنوان ژورنال:
  • CoRR

دوره abs/1310.7697  شماره 

صفحات  -

تاریخ انتشار 2013